TRACX2: a RAAM-like autoencoder modeling graded chunking in infant visual-sequence learning

نویسندگان

  • Robert M. French
  • Denis Mareschal
چکیده

Even newborn infants are able to extract structure from a stream of sensory inputs and yet, how this is achieved remains largely a mystery. We present a connectionist autoencoder model, TRACX2, that learns to extract sequence structure by gradually constructing chunks, storing these chunks in a distributed manner across its synaptic weights, and recognizing these chunks when they re-occur in the input stream. Chunks are graded rather than all-or-none in nature. As chunks are learned their component parts become more and more tightly bound together. TRACX2 successfully models the data from four experiments from the infant visual statistical-learning literature, including tasks involving lowsalience embedded chunk items, part-sequences, and illusory items. The model also captures performance differences across ages through the tuning of a single learning rate parameter. These results suggest that infant statistical learning is underpinned by the same domain general learning mechanism that operates in auditory statistical learning and, potentially, in adult artificial grammar learning. Introduction We live in a world in which events evolve over time. Consequently, our senses are bombarded with information that varies sequentially over time. One of the greatest challenges for cognition is to find structure within this stream of experiences. Even newborn infants are able to do this (Teinonen, et al. 2009; Bulf, Johnson & Valenza, 2011), and yet, how this is achieved remains largely a mystery. Two possibilities have been suggested (see Theissen, et al., 2013 for a detailed discussion). The first, characterised as statistical learning, involves using frequency and transition probabilities to construct an internal representation of the regularity boundaries among elements encountered. The second possibility suggests that elements that co-occur are recalled and simply grouped together – or chunked – into single units. Over time, these chunks can themselves be grouped into super-chunks or super-units. According to this view behaviour is determined by the recognition of these chunks stored in memory and associated with particular responses. What distinguishes these accounts 1 This article is an abridged, modified version of Mareschal, D. & French, R. M. (2017) TRACX2: a connectionist autoencoder using graded chunks to model infant visual statistical learning. Phil. Trans. R. Soc. B 2017 372 20160057; DOI: 10.1098/rstb.2016.0057. is that the former argues that it is the probabilistic structure of the input sequence that is represented and stored, whereas the later argues that specific cooccurring elements are stored, rather than the overarching statistical structure. Ample evidence in support of both of these views has been reported. We will argue that this is a false dichotomy: both transitional probability learning (statistical learning) and chunking co-exist in one system that smoothly transitions between these apparent modes of behaviour. The appearance of two modes of learning is an illusion because only a single mechanism underlies sequential learning; namely, Hebbian-style learning in a partially recurrent distributed neural network. Such a system encodes exemplars (typical of chunking mechanisms) while drawing on co-occurrence statistics (typical of statistical learning models). An important corollary of this approach is that chunks are graded in nature rather than all-or-none. Moreover, interference effects between chunks will follow a similarity gradient typical of other distributed neural network memory systems. Chunks are most frequently thought of as all-ornothing items. Who thinks of "cups" and "boards" when they see the word "cupboard"? Or "foot" and "ball" when they encounter the word "football"? Indeed, chunks like these have essentially the same status as "primitive" words like "boat" or "tree", which are not made of component sub-words. But new chunks do not suddenly appear ex nihilo in language. Rather, they are generally formed gradually, their component words becoming more and more bound together with time and usage. For example, when we encounter the words "smartphone", "carwash", or "petshop", we still clearly hear the component words. We hear them less in words like "sunburn" and "heartbeat". We hear them hardly at all in "automobile." How long did it take for people to stop hearing "auto" and "mobile" when they heard or read the word "automobile"? Like "automobile", it is likely that in a few years the current generation will no longer hear "smart" and "phone" when they hear the word "smartphone". This simple observation involving the graded nature of chunks is at the heart of the chunking mechanism in TRACX2. These ideas were implicit in our initial presentation of the TRACX model (French et al., 2011). In TRACX we showed that a connectionist autoencoder, augmented with conditional recurrence, could extract chunks from a stream of sequentially presented symbols. TRACX

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

TRACX2: a connectionist autoencoder using graded chunks to model infant visual statistical learning.

Even newborn infants are able to extract structure from a stream of sensory inputs; yet how this is achieved remains largely a mystery. We present a connectionist autoencoder model, TRACX2, that learns to extract sequence structure by gradually constructing chunks, storing these chunks in a distributed manner across its synaptic weights and recognizing these chunks when they re-occur in the inp...

متن کامل

Statistical and Chunking Processes in Adults' Visual Sequence Learning

Much research has documented learners’ ability to segment auditory and visual input into its component units. Two types of models have been designed to account for this phenomena: statistical models, in which learners represent statistical relations between elements, and chunking models, in which learners represent statistically coherent units of information. In a series of three experiments, w...

متن کامل

Linearizing Visual Processes with Convolutional Variational Autoencoders

This work studies the problem of modeling non-linear visual processes by learning linear generative models from observed sequences. We propose a joint learning framework, combining a Linear Dynamic System and a Variational Autoencoder with convolutional layers. After discussing several conditions for linearizing neural networks, we propose an architecture that allows Variational Autoencoders to...

متن کامل

Labeling Raam

In this paper we propose an extension of the Recursive Auto-Associative Memory (RAAM) by Pollack. This extension, the Labeling RAAM (LRAAM), is able to encode labeled graphs with cycles by representing pointers explicitly. A theoretical analysis of the constraints imposed on the weights by the learning task under the hypothesis of perfect learning and linear output units is presented. Cycles an...

متن کامل

Transduction Recursive Auto-Associative Memory: Learning Bilingual Compositional Distributed Vector Representations of Inversion Transduction Grammars

We introduce TRAAM, or Transduction RAAM, a fully bilingual generalization of Pollack’s (1990) monolingual Recursive Auto-Associative Memory neural network model, in which each distributed vector represents a bilingual constituent—i.e., an instance of a transduction rule, which specifies a relation between two monolingual constituents and how their subconstituents should be permuted. Bilingual ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017